Space-efficient Huffman codes revisited

نویسندگان

چکیده

A canonical Huffman code is an optimal prefix-free compression whose codewords enumerated in the lexicographical order form a list of binary words non-decreasing lengths. Gagie et al. (2015) gave representation this coding capable encoding and decoding symbol constant worst-case time. It uses σlg⁡ℓmax+o(σ)+O(ℓmax2) bits space, where σ ℓmax are alphabet size maximum codeword length, respectively. We refine their to reduce space complexity σlg⁡ℓmax(1+o(1)) while preserving encode decode times. Our algorithmic idea can be applied any code.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

An efficient decoding technique for Huffman codes

We present a new data structure for Huffman coding in which in addition to sending symbols in order of their appearance in the Huffman tree one needs to send codes of all circular leaf nodes (nodes with two adjacent external nodes), the number of which is always bounded above by half the number of symbols. We decode the text by using the memory efficient data structure proposed by Chen et al. [...

متن کامل

A Work Efficient Parallel Algorithm for Constructing Huffman Codes

Given an tilphabet C = (al! . . . , a,) and a corresponding list of weights [WI, . . . , w,], a Huffman code for this alphabet is a prefix code that minimizes the weighted length of a code string, defined to be Cr.., wili, where li is the length of the code assigned to ai. A. Huffman code can be generated in O(n log n) time for an unsorted list of weights alld in Cl(n) time if the weights are a...

متن کامل

Huffman Codes Lecturer : Michel

Shannon’s noiseless coding theorem tells us how compactly we can compress messages in which all letters are drawn independently from an alphabet A and we are given the probability pa of each letter a ∈ A appearing in the message. Shannon’s theorem says that, for random messages with n letters, the expected number of bits we need to transmit is at least nH(p) = −n ∑ a∈A pa log2 pa bits, and ther...

متن کامل

Entropy and Huffman Codes

We will show that • the entropy for a random variable gives a lower bound on the number of bits needed per character for a binary coding • Huffman codes are optimal in the average number of bits used per character among binary codes • the average bits per character used by Huffman codes is close to the entropy of the underlying random variable • one can get arbitrarily close to the entropy of a...

متن کامل

Discovery of Huffman codes

Large networks of IBM computers use it. So do high-definition television, modems and a popular electronic device that takes the brain work out of programming a videocassette recorder. All these digital wonders rely on the results of a 40-year-old term paper by a modest Massachusetts Institute of Technology graduate student—a data compression scheme known as Huffman encoding. In 1951 David A. Hu...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Information Processing Letters

سال: 2023

ISSN: ['1872-6119', '0020-0190']

DOI: https://doi.org/10.1016/j.ipl.2022.106274